Margin-Based Ranking Meets Boosting in the Middle

نویسندگان

  • Cynthia Rudin
  • Corinna Cortes
  • Mehryar Mohri
  • Robert E. Schapire
چکیده

We present a margin-based bound for ranking in a general setting, using the L∞ covering number of the hypothesis space as our complexity measure. Our bound suggests that ranking algorithms that maximize the ranking margin will generalize well. We produce a Smooth Margin Ranking algorithm, which is a modification of RankBoost analogous to Approximate Coordinate Ascent Boosting. We prove that this algorithm makes progress with respect to the ranking margin at every iteration and converges to a maximum margin solution. In the special case of bipartite ranking, the objective function of RankBoost is related to an exponentiated version of the AUC. In the empirical studies of Cortes and Mohri, and Caruana and Niculescu-Mizil, it has been observed that AdaBoost tends to maximize the AUC. In this paper, we give natural conditions such that AdaBoost maximizes the exponentiated loss associated with the AUC, i.e., conditions when AdaBoost and RankBoost will produce the same result, explaining the empirical observations.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Smooth Boosting for Margin-Based Ranking

We propose a new boosting algorithm for bipartite ranking problems. Our boosting algorithm, called SoftRankBoost, is a modification of RankBoost which maintains only smooth distributions over data. SoftRankBoost provably achieves approximately the maximum soft margin over all pairs of positive and negative examples, which implies high AUC score for future data.

متن کامل

Margin-based Ranking and an Equivalence between AdaBoost and RankBoost

We study boosting algorithms for learning to rank. We give a general margin-based bound for ranking based on covering numbers for the hypothesis space. Our bound suggests that algorithms that maximize the ranking margin will generalize well. We then describe a new algorithm, smooth margin ranking, that precisely converges to a maximum ranking-margin solution. The algorithm is a modification of ...

متن کامل

Multi-Resolution Cascades for Multiclass Object Detection

An algorithm for learning fast multiclass object detection cascades is introduced. It produces multi-resolution (MRes) cascades, whose early stages are binary target vs. non-target detectors that eliminate false positives, late stages multiclass classifiers that finely discriminate target classes, and middle stages have intermediate numbers of classes, determined in a data-driven manner. This M...

متن کامل

Calibrated Boosting-Forest

Excellent ranking power along with well calibrated probability estimates are needed in many classification tasks. In this paper, we introduce a technique, Calibrated Boosting-Forest1 that captures both. This novel technique is an ensemble of gradient boosting machines that can support both continuous and binary labels. While offering superior ranking power over any individual regression or clas...

متن کامل

New Multicategory Boosting Algorithms Based on Multicategory Fisher-consistent Losses.

Fisher-consistent loss functions play a fundamental role in the construction of successful binary margin-based classifiers. In this paper we establish the Fisher-consistency condition for multicategory classification problems. Our approach uses the margin vector concept which can be regarded as a multicategory generalization of the binary margin. We characterize a wide class of smooth convex lo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005